Goto

Collaborating Authors

 scientific machine


Geometry Matters: Benchmarking Scientific ML Approaches for Flow Prediction around Complex Geometries

Rabeh, Ali, Herron, Ethan, Balu, Aditya, Sarkar, Soumik, Hegde, Chinmay, Krishnamurthy, Adarsh, Ganapathysubramanian, Baskar

arXiv.org Artificial Intelligence

Rapid yet accurate simulations of fluid dynamics around complex geometries is critical in a variety of engineering and scientific applications, including aerodynamics and biomedical flows. However, while scientific machine learning (SciML) has shown promise, most studies are constrained to simple geometries, leaving complex, real-world scenarios underexplored. This study addresses this gap by benchmarking diverse SciML models, including neural operators and vision transformer-based foundation models, for fluid flow prediction over intricate geometries. Using a high-fidelity dataset of steady-state flows across various geometries, we evaluate the impact of geometric representations -- Signed Distance Fields (SDF) and binary masks -- on model accuracy, scalability, and generalization. Central to this effort is the introduction of a novel, unified scoring framework that integrates metrics for global accuracy, boundary layer fidelity, and physical consistency to enable a robust, comparative evaluation of model performance. Our findings demonstrate that foundation models significantly outperform neural operators, particularly in data-limited scenarios, and that SDF representations yield superior results with sufficient training data. Despite these advancements, all models struggle with out-of-distribution generalization, highlighting a critical challenge for future SciML applications. By advancing both evaluation methodologies and modeling capabilities, this work paves the way for robust and scalable ML solutions for fluid dynamics across complex geometries.


Federated scientific machine learning for approximating functions and solving differential equations with data heterogeneity

Zhang, Handi, Liu, Langchen, Lu, Lu

arXiv.org Artificial Intelligence

By leveraging neural networks, the emerging field of scientific machine learning (SciML) offers novel approaches to address complex problems governed by partial differential equations (PDEs). In practical applications, challenges arise due to the distributed essence of data, concerns about data privacy, or the impracticality of transferring large volumes of data. Federated learning (FL), a decentralized framework that enables the collaborative training of a global model while preserving data privacy, offers a solution to the challenges posed by isolated data pools and sensitive data issues. Here, this paper explores the integration of FL and SciML to approximate complex functions and solve differential equations. We propose two novel models: federated physics-informed neural networks (FedPINN) and federated deep operator networks (FedDeepONet). We further introduce various data generation methods to control the degree of non-independent and identically distributed (non-iid) data and utilize the 1-Wasserstein distance to quantify data heterogeneity in function approximation and PDE learning. We systematically investigate the relationship between data heterogeneity and federated model performance. Additionally, we propose a measure of weight divergence and develop a theoretical framework to establish growth bounds for weight divergence in federated learning compared to traditional centralized learning. To demonstrate the effectiveness of our methods, we conducted 10 experiments, including 2 on function approximation, 5 PDE problems on FedPINN, and 3 PDE problems on FedDeepONet. These experiments demonstrate that proposed federated methods surpass the models trained only using local data and achieve competitive accuracy of centralized models trained using all data.


Modeling chaotic Lorenz ODE System using Scientific Machine Learning

Kashyap, Sameera S, Dandekar, Raj Abhijit, Dandekar, Rajat, Panat, Sreedath

arXiv.org Artificial Intelligence

The Lorenz system of equations is a set of ordinary differential equations to represent a simplified model of atmospheric convection Sparrow [1982]. These set of equations have a wide range of applications in fields ranging from fluid mechanics to laser physics to weather prediction. One of the most interesting properties of the Lorenz ODE System is that it is chaotic in nature Fowler et al. [1982]. Small changes in the initial conditions can lead to vastly different outcomes in the end result Liao S. [2014]. When simulated over a given period, the Lorenz ODEs show oscillations in time. Usually, numerical methods implemented in computational software modeling tools like Python, Julia, or Matlab are used to simulate the Lorenz System of ODEs. These methods are inefficient as Lorentz equations are sensitive to initial conditions and minute changes to the conditions and tiny rounding errors can lead to the accumulation of numerical errors over time. Very few studies have been aimed at integrating machine learning-aided methods in simulating the chaotic Lorenz system. In this study, we provide a robust investigation of the effect of two physics-aided machine learning models in simulating the Lorenz system of ODEs: Neural Ordinary Differential Equations (Neural ODEs) Chen et al. [2018] and Universal Differential Equations (UDEs) Rackauckas et al. [2020a].


Leveraging Interpolation Models and Error Bounds for Verifiable Scientific Machine Learning

Chang, Tyler, Gillette, Andrew, Maulik, Romit

arXiv.org Machine Learning

Effective verification and validation techniques for modern scientific machine learning workflows are challenging to devise. Statistical methods are abundant and easily deployed, but often rely on speculative assumptions about the data and methods involved. Error bounds for classical interpolation techniques can provide mathematically rigorous estimates of accuracy, but often are difficult or impractical to determine computationally. In this work, we present a best-of-both-worlds approach to verifiable scientific machine learning by demonstrating that (1) multiple standard interpolation techniques have informative error bounds that can be computed or estimated efficiently; (2) comparative performance among distinct interpolants can aid in validation goals; (3) deploying interpolation methods on latent spaces generated by deep learning techniques enables some interpretability for black-box models. We present a detailed case study of our approach for predicting lift-drag ratios from airfoil images. Code developed for this work is available in a public Github repository.


Probabilistic Neural Networks (PNNs) for Modeling Aleatoric Uncertainty in Scientific Machine Learning

Pourkamali-Anaraki, Farhad, Husseini, Jamal F., Stapleton, Scott E.

arXiv.org Machine Learning

This paper investigates the use of probabilistic neural networks (PNNs) to model aleatoric uncertainty, which refers to the inherent variability in the input-output relationships of a system, often characterized by unequal variance or heteroscedasticity. Unlike traditional neural networks that produce deterministic outputs, PNNs generate probability distributions for the target variable, allowing the determination of both predicted means and intervals in regression scenarios. Contributions of this paper include the development of a probabilistic distance metric to optimize PNN architecture, and the deployment of PNNs in controlled data sets as well as a practical material science case involving fiber-reinforced composites. The findings confirm that PNNs effectively model aleatoric uncertainty, proving to be more appropriate than the commonly employed Gaussian process regression for this purpose. Specifically, in a real-world scientific machine learning context, PNNs yield remarkably accurate output mean estimates with R-squared scores approaching 0.97, and their predicted intervals exhibit a high correlation coefficient of nearly 0.80, closely matching observed data intervals. Hence, this research contributes to the ongoing exploration of leveraging the sophisticated representational capacity of neural networks to delineate complex input-output relationships in scientific problems.


Speeding up and reducing memory usage for scientific machine learning via mixed precision

Hayford, Joel, Goldman-Wetzler, Jacob, Wang, Eric, Lu, Lu

arXiv.org Artificial Intelligence

Scientific machine learning (SciML) has emerged as a versatile approach to address complex computational science and engineering problems. Within this field, physics-informed neural networks (PINNs) and deep operator networks (DeepONets) stand out as the leading techniques for solving partial differential equations by incorporating both physical equations and experimental data. However, training PINNs and DeepONets requires significant computational resources, including long computational times and large amounts of memory. In search of computational efficiency, training neural networks using half precision (float16) rather than the conventional single (float32) or double (float64) precision has gained substantial interest, given the inherent benefits of reduced computational time and memory consumed. However, we find that float16 cannot be applied to SciML methods, because of gradient divergence at the start of training, weight updates going to zero, and the inability to converge to a local minima. To overcome these limitations, we explore mixed precision, which is an approach that combines the float16 and float32 numerical formats to reduce memory usage and increase computational speed. Our experiments showcase that mixed precision training not only substantially decreases training times and memory demands but also maintains model accuracy. We also reinforce our empirical observations with a theoretical analysis. The research has broad implications for SciML in various computational applications.


Quantifying uncertainty for deep learning based forecasting and flow-reconstruction using neural architecture search ensembles

Maulik, Romit, Egele, Romain, Raghavan, Krishnan, Balaprakash, Prasanna

arXiv.org Artificial Intelligence

Classical problems in computational physics such as data-driven forecasting and signal reconstruction from sparse sensors have recently seen an explosion in deep neural network (DNN) based algorithmic approaches. However, most DNN models do not provide uncertainty estimates, which are crucial for establishing the trustworthiness of these techniques in downstream decision making tasks and scenarios. In recent years, ensemble-based methods have achieved significant success for the uncertainty quantification in DNNs on a number of benchmark problems. However, their performance on real-world applications remains under-explored. In this work, we present an automated approach to DNN discovery and demonstrate how this may also be utilized for ensemble-based uncertainty quantification. Specifically, we propose the use of a scalable neural and hyperparameter architecture search for discovering an ensemble of DNN models for complex dynamical systems. We highlight how the proposed method not only discovers high-performing neural network ensembles for our tasks, but also quantifies uncertainty seamlessly. This is achieved by using genetic algorithms and Bayesian optimization for sampling the search space of neural network architectures and hyperparameters. Subsequently, a model selection approach is used to identify candidate models for an ensemble set construction. Afterwards, a variance decomposition approach is used to estimate the uncertainty of the predictions from the ensemble. We demonstrate the feasibility of this framework for two tasks - forecasting from historical data and flow reconstruction from sparse sensors for the sea-surface temperature. We demonstrate superior performance from the ensemble in contrast with individual high-performing models and other benchmarks.


NSF-funded project to develop probabilistic scientific machine learning – TAMIDS Scientific Machine Learning Lab

#artificialintelligence

Across engineering and scientific disciplines, machine learning is the main method for analyzing and identifying patterns in big data and making informed decisions around that data. Recently, a new area within artificial intelligence called scientific machine learning has emerged, which introduces physics laws into machine learning models. Scientific machine learning combines the areas of artificial intelligence and scientific computation. Because scientific machine learning algorithms are informed and constrained by physics laws, they do not rely only on data and can even make predictions where there is no data. However, there has been little work on probabilistic methods in scientific machine learning, meaning that current algorithms cannot model uncertainty in the data or the physics.


Now Machine Learning Helps In Interpreting Battery Life

#artificialintelligence

A study carried out jointly by Stanford University, SLAC National Accelerator Laboratory, the Massachusetts Institute of Technology, and the Toyota Research Institute (TRI) demonstrated the use of machine learning algorithms to understand the lifecycle of lithium-ion batteries. Until now, machine learning in battery technology was limited to identifying patterns in data to speed up scientific analysis. The latest discovery will help researchers in designing and developing longer-lasting batteries. The research team has been working to develop a long-lasting electric vehicle battery that can be charged in 10 minutes. "Battery technology is important for any type of electric powertrain. By understanding the fundamental reactions that occur within the battery we can extend its life, enable faster charging and ultimately design better battery materials. We look forward to building on this work through future experiments to achieve lower-cost, better-performing batteries," said Patrick Herring, a senior scientist of Toyota Research Institute.


13 Data Science Things I Learned at JuliaCon 2020

#artificialintelligence

In this article, I will share 13 data science-related things I learned about Julia at JuliaCon 2020. I've grouped my learnings into 4 categories namely machine learning, tools, coding in Julia, and miscellaneous. MLJ.jl is a package from the Alan Turing Institute that serves as an interface to interact with machine learning algorithms in other packages. In addition, it provides functions to do common tasks in a machine learning project such as evaluating models, model stacking and hyperparameter tuning. There was an MLJ workshop run by Anthony Blaom, Thibaut Lienart, Geoffroy Dolphin, Okon Samuel, and Sebastian Vollmer where they demonstrate how to use MLJ to build models on the Iris dataset.